alt text

Generative AI for the Automated Design of Biomolecular Networks

Biomolecular networks underpin emerging technologies in synthetic biology—from robust biomanufacturing and metabolic engineering to smart therapeutics and cell-based diagnostics—and also provide a mechanistic language for understanding complex dynamics in natural and ecological systems. Yet designing chemical reaction networks (CRNs) that implement a desired dynamical function remains largely manual: while a proposed network can be checked by simulation, the reverse problem of discovering a network from a behavioral specification is difficult, requiring substantial human insight to navigate a vast space of topologies and kinetic parameters with nonlinear and possibly stochastic dynamics. Here we introduce GenAI-Net, a generative AI framework that automates CRN design by coupling an agent that proposes reactions to simulation-based evaluation defined by a user-specified objective. GenAI-Net efficiently produces novel, topologically diverse solutions across multiple design tasks, including dose responses, complex logic gates, classifiers, oscillators, and robust perfect adaptation in deterministic and stochastic settings (including noise reduction). By turning specifications into families of circuit candidates and reusable motifs, GenAI-Net provides a general route to programmable biomolecular circuit design and accelerates the translation from desired function to implementable mechanisms.

alt text

GenAI-Net is a framework for the automatic design of input--output chemical reaction networks (I/O CRNs). Users begin by selecting a desired design task (e.g. dose–response, oscillators, robust perfect adaptation, logic circuits, and classifiers). They then specify the intended (bio)chemical operating environment, including the set of chemical species that may appear in the network, the designation of inputs (e.g., externally set signals u1) and outputs (regulated/readout species), and any contextual constraints implied by the application setting (e.g. in vitro, cellular contexts, or others). Finally, users choose an appropriate kinetic model class (e.g., mass-action or Michaelis–Menten) and provide (or select) an available reaction library containing M permissible reactions from which candidate networks may be assembled. Given these specifications, GenAI-Net iteratively generates candidate I/O CRNs, simulates their dynamics under the chosen kinetics, and evaluates performance against the task objective using quantitative metrics (e.g. time-domain error and frequency-domain/Fourier features). These evaluations provide a learning signal used to train the agent, closing the design loop: the agent proposes new networks conditioned on prior performance, progressively improving the generated candidates. The output of a design run is a batch of I/O CRNs sampled from the learned search policy, enabling downstream selection, analysis, and implementation. GenAI-Net returns multiple top-ranked candidate I/O CRNs, each represented by a specific reaction topology, parameters, and its corresponding predicted input--output behavior. Diversity graphs can used to monitor the topological diversity of the generated I/O CRNs. The example to the right highlights how topologically distinct candidate networks can realize similar target specifications (e.g., matching a desired dose–response curve) while differing in internal reaction structure, enabling users to trade off performance, simplicity, and implementability when choosing a final design.

Machine Learning in Music

Identifying composers from their music has traditionally been a skill reserved for trained musical ears, yet it now poses an intriguing challenge for machine learning and artificial intelligence. Classical music composers each possess a distinctive style, influenced by factors such as rhythmic structure and pitch which affect the speed and mood (be it joyful, passionate, or dramatic) of a piece. These stylistic elements often mirror the composer's lifestyle or the era they lived in. Our research is focused on uncovering the specific structures in piano compositions that define a composer’s style, thus facilitating automatic composer recognition.

In our research, we utilize MIDI files as inputs for our composer recognition algorithm. The initial step involves extracting features from these MIDI files, with a focus on n-grams, a technique widely used in natural language processing and statistical language modeling. n-grams are sequences of 'n' items, which in our context might be syllables, letters, or words, depending on the specific application. To illustrate, consider the sentence: "The cat sat on the mat and the cat sat on the chair." From this, we can form 3-grams such as "the cat sat", "cat sat on", "sat on the", "on the mat" and so on. Below is a table summarizing the occurrences of each 3-gram extracted from the sentence.

3-grams Occurrences
The cat sat 2
cat sat on 2
sat on the 2
on the mat 1
the mat and 1
mat and the 1
and the cat 1
on the chair 1

We utilize a similar method to create music n-grams that focus on both the melody and bass, considering their pitch variations and rhythmic structures. Initially, preprocessing is carried out on the MIDI files to separate the melody from the bass before constructing the n-grams.

alt text

Here's a simple example of how music n-grams are formed from a piece of the score of the Lebanese National Anthem:

alt text

We employ a cortical algorithm for feature reduction and classification, conducting our experiments on a database compiled from the Humdrum project library. The MIDI files are processed using a MATLAB MIDI toolbox. Our approach has been tested for classifying unknown composers and identifying distinct musical styles, achieving a recognition rate of 94.4% on our custom database of 1,197 pieces. This success was attained using only 0.1% of the 231,542 features generated, emphasizing the efficacy of the feature reduction procedure.

After examining the most significant features that led to the best performance, it was possible to identify five key features as most critical: the occurrence of notes with onset on quarter beats, complete beats, and half beats, along with two n-gram combinations of bass notes. From a musical perspective, these findings suggest that piano composers primarily distinguish their style through the handling of bass notes and rhythmic onsets. While melodies are crucial in defining a composer's style, our results indicate that the arrangement of notes across beats and the development of bass patterns in conjunction with the melody provide deeper insights into a composer’s stylistic signature. This is notable since humans typically focus more on high-pitched melodies than on bass notes, yet composers uniquely define their style through variations in rhythm and bass lines accompanying the same melody. Further analysis through one-to-one composer classification revealed confusion rates varying significantly, from a high of 17.1% between Mozart and Haydn to as low as 0.5% between Vivaldi and Joplin, with Joplin being distinctly recognized due to his unique era of composition. The challenges in distinguishing between Haydn and Mozart, as well as Beethoven and Chopin, underscore the subtle stylistic similarities shared among these composers, corroborating historical accounts of their influences and stylistic overlaps.

References: